Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher.
Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?
Some links on this page may take you to non-federal websites. Their policies may differ from this site.
-
Accurate in-season prediction of seed yield and seed composition traits such as oil and protein are useful for gaining accuracy and efficiency in soybean breeding. These predictions can also inform farmers, enabling them to improve their field management practices, and guide their market decisions. We report a Transformer-based deep learning framework built on 30 years of multi-environment performance data from the Northern and Southern Uniform Soybean Tests (UST) across North America. Unlike earlier studies on seed yield, oil and protein prediction that focus on limited years, regions, single modalities, we utilized a comprehensive dataset that includes weather, genotype, and management factors, ensuring a more holistic approach to soybean yield, oil, and protein prediction. Our model integrates multivariate time-series weather data with genotypic relationship information, maturity group, and geographic location, to predict variety performance in diverse environments. Our model captures complex temporal patterns associated with trait variability; showing high predictive accuracy (R2) of 77.6 ± 0.2%, 63.9 ± 4.7%, and 79.3 ± 2.3% for seed yield, oil, and protein, respectively. Additionally, for seed yield, we also evaluated multiple interpretability methods to assess feature importance for predictor variables and critical growing timepoints, and solar radiation and temperature were noted as the key predictors. Overall, these results demonstrate the usefulness of a Transformer-based model in trait predictions, and the utility of large cooperative datasets from breeding programs.more » « less
-
The performance of 3D reconstruction using Neural Radiance Fields (NeRFs) for outdoor phenotyping of plants is strongly influenced by the imaging modality used for data collection. We compare drone, handheld, and 360° ground robot datasets collected over soybean and mungbean plots, and evaluate reconstruction quality using 2D metrics PSNR, SSIM, LPIPS, and 3D geometric metrics precision, recall, and F1 score. Drone imagery produced the highest geometric fidelity, handheld captures achieved the strongest 2D appearance quality, and the 360° captures lagged in both metrics due to spherical distortion and motion artifacts. The consistency of the drone-based reconstructions highlights its suitability for field-scale 3D modeling and positions it as a practical foundation for future phenotyping applicationsmore » « less
An official website of the United States government

Full Text Available